DyNet: The Dynamic Neural Network Toolkit

نویسندگان

  • Graham Neubig
  • Chris Dyer
  • Yoav Goldberg
  • Austin Matthews
  • Waleed Ammar
  • Antonios Anastasopoulos
  • Miguel Ballesteros
  • David Chiang
  • Daniel Clothiaux
  • Trevor Cohn
  • Kevin Duh
  • Manaal Faruqui
  • Cynthia Gan
  • Dan Garrette
  • Yangfeng Ji
  • Lingpeng Kong
  • Adhiguna Kuncoro
  • Gaurav Kumar
  • Chaitanya Malaviya
  • Paul Michel
  • Yusuke Oda
  • Matthew Richardson
  • Naomi Saphra
  • Swabha Swayamdipta
  • Pengcheng Yin
چکیده

We describe DyNet, a toolkit for implementing neural network models based on dynamic declaration of network structure. In the static declaration strategy that is used in toolkits like Theano, CNTK, and TensorFlow, the user first defines a computation graph (a symbolic representation of the computation), and then examples are fed into an engine that executes this computation and computes its derivatives. In DyNet’s dynamic declaration strategy, computation graph construction is mostly transparent, being implicitly constructed by executing procedural code that computes the network outputs, and the user is free to use different network structures for each input. Dynamic declaration thus facilitates the implementation of more complicated network architectures, and DyNet is specifically designed to allow users to implement their models in a way that is idiomatic in their preferred programming language (C++ or Python). One challenge with dynamic declaration is that because the symbolic computation graph is defined anew for every training example, its construction must have low overhead. To achieve this, DyNet has an optimized C++ backend and lightweight graph representation. Experiments show that DyNet’s speeds are faster than or comparable with static declaration toolkits, and significantly faster than Chainer, another dynamic declaration toolkit. DyNet is released opensource under the Apache 2.0 license, and available at http://github.com/clab/dynet. Carnegie Mellon University, Pittsburgh, PA, USA Nara Institute of Science and Technology, Ikoma, Japan DeepMind, London, UK Bar Ilan University, Ramat Gan, Israel Allen Institute for Artificial Intelligence, Seattle, WA, USA University of Notre Dame, Notre Dame, IN, USA IBM T.J. Watson Research Center, Yorktown Heights, NY, USA University of Melbourne, Melbourne, Australia Johns Hopkins University, Baltimore, MD, USA Google, New York, NY, USA Google, Mountain View, CA, USA University of Washington, Seattle, USA Microsoft Research, Redmond, WA, USA University of Edinburgh, Edinburgh, UK 1 ar X iv :1 70 1. 03 98 0v 1 [ st at .M L ] 1 5 Ja n 20 17

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

On-the-fly Operation Batching in Dynamic Computation Graphs

Dynamic neural network toolkits such as PyTorch, DyNet, and Chainer offer more flexibility for implementing models that cope with data of varying dimensions and structure, relative to toolkits that operate on statically declared computations (e.g., TensorFlow, CNTK, and Theano). However, existing toolkits—both static and dynamic—require that the developer organize the computations into the batc...

متن کامل

DyNet: visualization and analysis of dynamic molecular interaction networks

UNLABELLED : The ability to experimentally determine molecular interactions on an almost proteome-wide scale under different conditions is enabling researchers to move from static to dynamic network analysis, uncovering new insights into how interaction networks are physically rewired in response to different stimuli and in disease. Dynamic interaction data presents a special challenge in netwo...

متن کامل

Cavs: A Vertex-centric Programming Interface for Dynamic Neural Networks

Recent deep learning (DL) models have moved beyond static network architectures to dynamic ones, handling data where the network structure changes every example, such as sequences of variable lengths, trees, and graphs. Existing dataflow-based programming models for DL—both static and dynamic declaration—either cannot readily express these dynamic models, or are inefficient due to repeated data...

متن کامل

Comparative Study of Static and Dynamic Artificial Neural Network Models in Forecasting of Tehran Stock Exchange

During the recent decades, neural network models have been focused upon by researchers due to their more real performance and on this basis, different types of these models have been used in forecasting. Now, there is a question that which kind of these models has more explanatory power in forecasting the future processes of the stock. In line with this, the present paper made a comparison betw...

متن کامل

A Convolutional Neural Network based on Adaptive Pooling for Classification of Noisy Images

Convolutional neural network is one of the effective methods for classifying images that performs learning using convolutional, pooling and fully-connected layers. All kinds of noise disrupt the operation of this network. Noise images reduce classification accuracy and increase convolutional neural network training time. Noise is an unwanted signal that destroys the original signal. Noise chang...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:
  • CoRR

دوره abs/1701.03980  شماره 

صفحات  -

تاریخ انتشار 2017